671 research outputs found

    Finding suitable activity clusters for decomposed process discovery

    Get PDF
    Event data can be found in any information system and provide the starting point for a range of process mining techniques. The widespread availability of large amounts of event data also creates new challenges. Existing process mining techniques are often unable to handle "big event data" adequately. Decomposed process mining aims to solve this problem by decomposing the process mining problem into many smaller problems which can be solved in less time, using less resources, or even in parallel. Many decomposed process mining techniques have been proposed in literature. Analysis shows that even though the decomposition step takes a relatively small amount of time, it is of key importance in finding a high-quality process model and for the computation time required to discover the individual parts. Currently there is no way to assess the quality of a decomposition beforehand. We define three quality notions that can be used to assess a decomposition, before using it to discover a model or check conformance with. We then propose a decomposition approach that uses these notions and is able to find a high-quality decomposition in little time. Keywords: decomposed process mining, decomposed process discovery, distributed computing, event lo

    Discovering duplicate tasks in transition systems for the simplification of process models

    Get PDF
    This work presents a set of methods to improve the understandability of process models. Traditionally, simplification methods trade off quality metrics, such as fitness or precision. Conversely, the methods proposed in this paper produce simplified models while preserving or even increasing fidelity metrics. The first problem addressed in the paper is the discovery of duplicate tasks. A new method is proposed that avoids overfitting by working on the transition system generated by the log. The method is able to discover duplicate tasks even in the presence of concurrency and choice. The second problem is the structural simplification of the model by identifying optional and repetitive tasks. The tasks are substituted by annotated events that allow the removal of silent tasks and reduce the complexity of the model. An important feature of the methods proposed in this paper is that they are independent from the actual miner used for process discovery.Peer ReviewedPostprint (author's final draft

    A Language for the Specification of Administrative Workflow Processes with Emphasis on Actors’ Views

    Get PDF
    International audienceAdministrative workflows refer to variable business processes in which all cases are known; tasks are predictable and their sequencing rules are simple and clearly defined. When such processes are collaboratively executed by several actors, it may be desirable, for security reasons (confidentiality), that each of them has at all times, only a partial perception (this is what we call "actor's view") of the current process state. This concern seems sufficiently important to be considered when specifying such workflows. However, traditional workflow specification languages (BPMN, BPEL, YAWL) only partially address it. This is why we present in this paper, a new language for specifying administrative workflows that allows us not only to simply model all of the processes tasks and their sequence, but also and especially to explicitly express the rights of the various actors with respect to each of them, in order to guarantee a certain degree of security. The proposed model is an executable grammatical specification that allows to express using decorated productions, the different types of basic flows (sequential, parallel, alternative and iterative) that are found in workflow specification languages; moreover, it also allows to specify the rights of each actor in each process and on its data in a formalism similar to that used in UNIX-like operating systems

    A Method to Improve the Early Stages of the Robotic Process Automation Lifecycle

    Get PDF
    The robotic automation of processes is of much interest to organizations. A common use case is to automate the repetitive manual tasks (or processes) that are currently done by back-office staff through some information system (IS). The lifecycle of any Robotic Process Automation (RPA) project starts with the analysis of the process to automate. This is a very time-consuming phase, which in practical settings often relies on the study of process documentation. Such documentation is typically incomplete or inaccurate, e.g., some documented cases never occur, occurring cases are not documented, or documented cases differ from reality. To deploy robots in a production environment that are designed on such a shaky basis entails a high risk. This paper describes and evaluates a new proposal for the early stages of an RPA project: the analysis of a process and its subsequent design. The idea is to leverage the knowledge of back-office staff, which starts by monitoring them in a non-invasive manner. This is done through a screen-mousekey- logger, i.e., a sequence of images, mouse actions, and key actions are stored along with their timestamps. The log which is obtained in this way is transformed into a UI log through image-analysis techniques (e.g., fingerprinting or OCR) and then transformed into a process model by the use of process discovery algorithms. We evaluated this method for two real-life, industrial cases. The evaluation shows clear and substantial benefits in terms of accuracy and speed. This paper presents the method, along with a number of limitations that need to be addressed such that it can be applied in wider contexts.Ministerio de Economía y Competitividad TIN2016-76956-C3-2-

    The Relational Process Structure

    Get PDF
    Using data-centric process paradigms, small processes such as artifacts, object lifecycles, or Proclets have become an alternative to large, monolithic models. In these paradigms, a business process arises from the interactions between small processes. However, many-to-many relationships may exist between different process types, requiring careful consideration to ensure that the interactions between processes can be purposefully coordinated. Although several concepts exist for modeling interrelated processes, a concept that considers both many-to-many relationships and cardinality constraints is missing. Furthermore, existing concepts focus on design-time, neglecting the complexity introduced by many-to-many relationships when enacting extensive process structures at run-time. The knowledge which process instances are related to which other process instances is essential. This paper proposes the relational process structure, a concept providing full support for many-to-many-relationships and cardinality constraints at both design- and run-time. The relational process structure represents a cornerstone to the proper coordination of interrelated processes

    Learning Hybrid Process Models From Events: Process Discovery Without Faking Confidence

    Full text link
    Process discovery techniques return process models that are either formal (precisely describing the possible behaviors) or informal (merely a "picture" not allowing for any form of formal reasoning). Formal models are able to classify traces (i.e., sequences of events) as fitting or non-fitting. Most process mining approaches described in the literature produce such models. This is in stark contrast with the over 25 available commercial process mining tools that only discover informal process models that remain deliberately vague on the precise set of possible traces. There are two main reasons why vendors resort to such models: scalability and simplicity. In this paper, we propose to combine the best of both worlds: discovering hybrid process models that have formal and informal elements. As a proof of concept we present a discovery technique based on hybrid Petri nets. These models allow for formal reasoning, but also reveal information that cannot be captured in mainstream formal models. A novel discovery algorithm returning hybrid Petri nets has been implemented in ProM and has been applied to several real-life event logs. The results clearly demonstrate the advantages of remaining "vague" when there is not enough "evidence" in the data or standard modeling constructs do not "fit". Moreover, the approach is scalable enough to be incorporated in industrial-strength process mining tools.Comment: 25 pages, 12 figure

    Anti-alignments in conformance checking: the dark side of process models

    Get PDF
    Conformance checking techniques asses the suitability of a process model in representing an underlying process, observed through a collection of real executions. These techniques suffer from the wellknown state space explosion problem, hence handling process models exhibiting large or even infinite state spaces remains a challenge. One important metric in conformance checking is to asses the precision of the model with respect to the observed executions, i.e., characterize the ability of the model to produce behavior unrelated to the one observed. By avoiding the computation of the full state space of a model, current techniques only provide estimations of the precision metric, which in some situations tend to be very optimistic, thus hiding real problems a process model may have. In this paper we present the notion of antialignment as a concept to help unveiling traces in the model that may deviate significantly from the observed behavior. Using anti-alignments, current estimations can be improved, e.g., in precision checking. We show how to express the problem of finding anti-alignments as the satisfiability of a Boolean formula, and provide a tool which can deal with large models efficiently.Peer ReviewedPostprint (author's final draft
    • …
    corecore